Investigation of Pre-Trained Bidirectional Encoder Representations from Transformers Checkpoints for Indonesian Abstractive Text Summarization
نویسندگان
چکیده
Text summarization aims to reduce text by removing less useful information obtain quickly and precisely. In Indonesian abstractive summarization, the research mostly focuses on multi-document which methods will not work optimally in single-document summarization. As public datasets works English are focusing this study emphasized Abstractive studies frequently use Bidirectional Encoder Representations from Transformers (BERT), since BERT checkpoint is available, it was employed study. This investigated of IndoSum dataset using BERTSum model. The investigation proceeded various combinations model encoders, embedding sizes, decoders. Evaluation results showed that models with more size used Generative Pre-Training (GPT)-like decoder could improve Recall-Oriented Understudy for Gisting (ROUGE) score BERTScore results.
منابع مشابه
Text Generation for Abstractive Summarization
We have begun work on a framework for abstractive summarization and decided to focus on a module for text generation. For TAC 2010, we thus move away from sentence extraction. Each sentence in the summary we generate is based on a document sentence but it usually contains a smaller amount of information and uses fewer words. The system uses the output of a syntactic parser for a sentence and th...
متن کاملNeural Abstractive Text Summarization
Abstractive text summarization is a complex task whose goal is to generate a concise version of a text without necessarily reusing the sentences from the original source, but still preserving the meaning and the key contents. We address this issue by modeling the problem as a sequence to sequence learning and exploiting Recurrent Neural Networks (RNNs). This work is a discussion about our ongoi...
متن کاملDecoupling Encoder and Decoder Networks for Abstractive Document Summarization
Abstractive document summarization seeks to automatically generate a summary for a document, based on some abstract “understanding” of the original document. State-of-the-art techniques traditionally use attentive encoder–decoder architectures. However, due to the large number of parameters in these models, they require large training datasets and long training times. In this paper, we propose ...
متن کاملGenerative Adversarial Network for Abstractive Text Summarization
In this paper, we propose an adversarial process for abstractive text summarization, in which we simultaneously train a generative model G and a discriminative model D. In particular, we build the generator G as an agent of reinforcement learning, which takes the raw text as input and predicts the abstractive summarization. We also build a discriminator which attempts to distinguish the generat...
متن کاملFramework for Abstractive Summarization using Text-to-Text Generation
We propose a new, ambitious framework for abstractive summarization, which aims at selecting the content of a summary not from sentences, but from an abstract representation of the source documents. This abstract representation relies on the concept of Information Items (INIT), which we define as the smallest element of coherent information in a text or a sentence. Our framework differs from pr...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Journal of ICT
سال: 2021
ISSN: ['1675-414X', '2180-3862']
DOI: https://doi.org/10.32890/jict2022.21.1.4